Research
Security News
Malicious npm Packages Inject SSH Backdoors via Typosquatted Libraries
Socket’s threat research team has detected six malicious npm packages typosquatting popular libraries to insert SSH backdoors.
Lightweight and real-time data functional stream programming framework like event-stream, written in ES6 using async await with multi-threading and typescript support
Version 4
Scramjet is a fast, simple, functional reactive stream programming framework written on top of node.js object streams. The code is written by chaining functions that transform the streamed data, including well known map, filter and reduce and fully compatible with ES7 async/await. Thanks to it some built in optimizations scramjet is much faster and much much simpler than similar frameworks when using asynchronous operations.
The main advantage of scramjet is running asynchronous operations on your data streams. First of all it allows you to perform the transformations both synchronously and asynchronously by using the same API - so now you can "map" your stream from whatever source and call any number of API's consecutively. And if you're after some heavy maths there's an option of running your stream as multi-threaded!
We are working on the next version of Scramjet Framework and are very eager for your feedback! You can see and test pre-v5
:
How about a full API to API migration, reading a long list of items from one API and checking them one after another, pushing them to another API? With simultaneous request control? And outputting the log of the conversion? Easy!
const fetch = require("node-fetch");
const get = async (url, options = {}) => (await fetch(url, options)).json;
const { StringStream } = require("scramjet");
StringStream.from( // fetch your API to a scramjet stream
() => get("https://api.example.org/v1/shows/list")
)
.setOptions({maxParallel: 4}) // set your options
.lines() // split the stream by line
.parse(line => { // parse strings to data
const [id, title, url] = line.split(",");
return { id, title, url };
})
.map(async myShow => get({ // use asynchronous mapping (for example send requests)
uri: `http://api.local/set/${myShow.id}`,
body: JSON.stringify(myShow)
}))
.stringify(resp => `+ Updated "${resp}"`)
.catch(err => `! Error occured ${err.uri}`) // handle errors
.append("\n")
.pipe(process.stdout) // use any stream
;
Here you can find a most basic guide on how to execute the above example starting from just having access to some command line: Scramjet from Scratch
You can now run stream processing programs with our Scramjet Transform Hub. It will allow you to deploy and execute programs on local and remote environments of your choice and it's as easy as:
npm i -g @scramjet/sth @scramjet/cli
scramjet-transform-hub &
si run <path-to-your-program-dir>
See more info:
Scramjet uses functional programming to run transformations on your data streams in a fashion very similar to the well known event-stream node module. First create a stream from a source:
Use DataStream.from(someThing)
to create a new stream from an Array, Generator, AsyncGenerator,
Iterator or Readable stream. See the DataStream.from docs
for more information, here's a sample.
/* global StringStream, fs */
StringStream
.from(fs.createReadStream("./log.txt")) // get from any readable stream
.lines() // split the stream by line
.use("./your-file") // use some transforms from another file
;
Use DataStream.pipeline(readable, transforms)
to create a pipeline of transform streams and/or
stream modules. Any number of consecutive arguments will get piped one into another.
/* global StringStream, fs, gzip */
StringStream
.pipeline( // process a number of streams
fs.createReadStream("./log.txt.gz"),
gzip.unzip() // all errors here will get forwarded
)
.lines() // split the stream by line
.use("./your-file") // use some transforms from another file
;
Some methods like from
, use
, flatMap
allow using ES6 generators and ES7 async generators:
const fetch = require("node-fetch");
const { StringStream } = require("scramjet");
StringStream
.from(
async function* () { // construct a stream from an async generator
yield "houses\n"; // yield - push a stream chunk
// yield - push a whole stream
yield* (await fetch("https://example.org/categories")).body;
},
{maxParallel: 4} // set your options
)
.lines() // split the stream by line
.flatMap(async function* (category) {
const req = await fetch(`https://example.org/posts/${category}/`);
yield* await req.json(); // yield - push a whole array
})
.catch(err => `! Error occured ${err.uri}`)
.toStringStream()
.append("\n")
.pipe(process.stdout) // pipe to any output
;
Most transformations are done by passing a transform function. You can write your function in three ways:
Example: a simple stream transform that outputs a stream of objects of the same id property and the length of the value string.
DataStream
.from(items)
.map(
(item) => ({id: item.id, length: item.value.length})
)
Example: A simple stream that uses Fetch API
to get all the contents of all entries in the stream
StringStream
.from(urls)
.map(
async (url) => fetch(url).then(res => res.json())
)
.JSONParse()
Example: A simple stream that fetches an url mentioned in the incoming object
datastream.map(
(item) => new Promise((resolve, reject) => {
request(item.url, (err, res, data) => {
if (err)
reject(err); // will emit an "error" event on the stream
else
resolve(data);
});
})
)
The actual logic of this transform function is as if you passed your function to the then
method of a Promise
resolved with the data from the input stream.
To distribute your code among the processor cores, just use the method distribute
:
datastream.distribute(
16, // number of threads
(stream) => {
// multi-threaded code goes here.
// it MUST return a valid stream back to the main thread.
}
)
Scramjet allows writing simple modules that are resolved in the same way as node's require
. A module
is a simple javascript file that exposes a function taking a stream and any number of following arguments
as default export.
Here's an example:
module.exports = (stream, arg1) => {
const mapper = (chunk) => mapper(chunk, arg1);
return stream.map(mapper);
}
Then it can be used with DataStream.use
function like this:
myStream.use("./path/to/my-module", "arg1");
If these modules are published you can also simply use myStream.use("published-module")
.
For more universal modules you can use helper methods createTransformModule
and createReadModule
that scramjet exports. See more in about this in this blog post Scramjet Modules.
Scramjet aims to be fully documented and expose TypeScript declarations. First version to include definitions in .d.ts
folder is 4.15.0. More TypeScript support will be added with next versions, so feel free to report issues in GitHub.
Here's the list of the exposed classes and methods, please review the specific documentation for details:
scramjet
- module exports explainedscramjet.DataStream
- the base class for all scramjet classes, object stream.scramjet.BufferStream
- a stream of Buffers.scramjet.StringStream
- a stream of Strings.scramjet.NumberStream
- a stream of Numbersscramjet.MultiStream
- A group of streams (for multi-threading and muxing).Note that:
Check out the command line interface for simplified scramjet usage with scramjet-cli
$ sjr -i http://datasource.org/file.csv ./transform-module-1 ./transform-module-1 | gzip > logs.gz
DataStream is the primary stream type for Scramjet. When you parse your stream, just pipe it you can then perform calculations on the data objects streamed through your flow.
Use as:
const { DataStream } = require('scramjet');
await (DataStream.from(aStream) // create a DataStream
.map(findInFiles) // read some data asynchronously
.map(sendToAPI) // send the data somewhere
.run()); // wait until end
Detailed :DataStream docs here
Most popular methods:
new DataStream([opts])
- Create the DataStream.dataStream.map(func, [ClassType]) ↺
- Transforms stream objects into new ones, just like Array.prototype.mapdataStream.filter(func) ↺
- Filters object based on the function outcome, just like Array.prototype.filter.dataStream.reduce(func, into) ⇄
- Reduces the stream into a given accumulatordataStream.do(func) ↺
- Perform an asynchronous operation without changing or resuming the stream.dataStream.all(functions) ↺
- Processes a number of functions in parallel, returns a stream of arrays of results.dataStream.race(functions) ↺
- Processes a number of functions in parallel, returns the first resolved.dataStream.unorder(func)
- Allows processing items without keeping orderdataStream.into(func, into) ↺
- Allows own implementation of stream chaining.dataStream.use(func) ↺
- Calls the passed method in place with the stream as first argument, returns result.dataStream.run() ⇄
- Consumes all stream items doing nothing. Resolves when the stream is ended.dataStream.tap() ↺
- Stops merging transform Functions at the current place in the command chain.dataStream.whenRead() ⇄
- Reads a chunk from the stream and resolves the promise when read.dataStream.whenWrote(chunk) ⇄
- Writes a chunk to the stream and returns a Promise resolved when more chunks can be written.dataStream.whenEnd() ⇄
- Resolves when stream ends - rejects on uncaught errordataStream.whenDrained() ⇄
- Returns a promise that resolves when the stream is draineddataStream.whenError() ⇄
- Returns a promise that resolves (!) when the stream is errorsdataStream.setOptions(options) ↺
- Allows resetting stream options.dataStream.copy(func) ↺
- Returns a copy of the streamdataStream.tee(func) ↺
- Duplicate the streamdataStream.each(func) ↺
- Performs an operation on every chunk, without changing the streamdataStream.while(func) ↺
- Reads the stream while the function outcome is truthy.dataStream.until(func) ↺
- Reads the stream until the function outcome is truthy.dataStream.catch(callback) ↺
- Provides a way to catch errors in chained streams.dataStream.raise(err) ⇄
- Executes all error handlers and if none resolves, then emits an error.dataStream.bufferify(serializer) : BufferStream ↺
- Creates a BufferStream.dataStream.stringify([serializer]) : StringStream ↺
- Creates a StringStream.dataStream.toArray([initial]) : Array.<any> ⇄
- Aggregates the stream into a single ArraydataStream.toGenerator() : Generator.<Promise.<any>>
- Returns an async generatordataStream.pull(pullable) : Promise.<any> ⇄
- Pulls in any readable stream, resolves when the pulled stream ends.dataStream.shift(count, func) ↺
- Shifts the first n items from the stream and pushes out the remaining ones.dataStream.peek(count, func) ↺
- Allows previewing some of the streams data without removing them from the stream.dataStream.slice([start], [length]) ↺
- Slices out a part of the stream to the passed Function.dataStream.assign(func) ↺
- Transforms stream objects by assigning the properties from the returneddataStream.empty(callback) ↺
- Called only before the stream ends without passing any itemsdataStream.unshift() ↺
- Pushes any data at call time (essentially at the beginning of the stream)dataStream.endWith(item) ↺
- Pushes any data at end of streamdataStream.accumulate(func, into) : Promise.<any> ⇄
- Accumulates data into the object.~~dataStream.consume(func) ⇄~~
- Consumes the stream by running each FunctiondataStream.reduceNow(func, into) : * ↺
- Reduces the stream into the given object, returning it immediately.dataStream.remap(func, [ClassType]) ↺
- Remaps the stream into a new stream.dataStream.flatMap(func, [ClassType]) ↺
- Takes any method that returns any iterable and flattens the result.dataStream.flatten() : DataStream ↺
- A shorthand for streams of arrays or iterables to flatten them.dataStream.concat() ↺
- Returns a new stream that will append the passed streams to the calleedataStream.join(item) ↺
- Method will put the passed object between items. It can also be a function call or generator / iterator.dataStream.keep([count]) ↺
- Keep a buffer of n-chunks for use with {@see DataStream..rewind}dataStream.rewind([count]) ↺
- Rewinds the buffered chunks the specified length backwards. Requires a prior call to {@see DataStream..keep}dataStream.stack([count], [drop]) ↺
- Returns a stream that stacks up incoming items always feeding out the newest items first.dataStream.distribute([affinity], [clusterFunc], [options]) ↺
- Distributes processing into multiple sub-processes or threads if you like.dataStream.separateInto(streams, affinity) ↺
- Separates stream into a hash of streams. Does not create new streams!dataStream.separate(affinity, [createOptions], [ClassType]) : MultiStream ↺
- Separates execution to multiple streams using the hashes returned by the passed Function.dataStream.delegate(delegateFunc, worker, [plugins]) ↺
- Delegates work to a specified worker.dataStream.rate(cps, [options]) ↺
- Limit the rate of the stream to a given number of chunks per second or given timeframe.dataStream.batch(count) ↺
- Aggregates chunks in arrays given number of number of items long.dataStream.timeBatch(ms, [count]) ↺
- Aggregates chunks to arrays not delaying output by more than the given number of ms.dataStream.nagle([size], [ms]) ↺
- Performs the Nagle's algorithm on the data. In essence it waits until we receive some more data and releases themdataStream.window(length) : WindowStream ↺
- Returns a WindowStream of the specified lengthdataStream.toJSONArray([enclosure]) : StringStream ↺
- Transforms the stream to a streamed JSON array.dataStream.toJSONObject([entryCallback], [enclosure]) : StringStream ↺
- Transforms the stream to a streamed JSON object.dataStream.JSONStringify([endline]) : StringStream ↺
- Returns a StringStream containing JSON per item with optional end linedataStream.CSVStringify([options]) : StringStream ↺
- Stringifies CSV to DataString using 'papaparse' module.dataStream.exec(command, [options])
- Executes a given sub-process with arguments and pipes the current stream into it while returning the output as another DataStream.dataStream.debug(func) : DataStream ↺
- Injects a debugger
statement when called.dataStream.toBufferStream(serializer) : BufferStream ↺
- Creates a BufferStream.dataStream.toStringStream([serializer]) : StringStream ↺
- Creates a StringStream.dataStream.toBufferStream(serializer) : BufferStream ↺
- Creates a BufferStream.dataStream.toStringStream([serializer]) : StringStream ↺
- Creates a StringStream.DataStream:from(input, [options]) : DataStream
- Returns a DataStream from pretty much anything sensibly possible.DataStream:pipeline(readable) : DataStream
- Creates a pipeline of streams and returns a scramjet stream.DataStream:fromArray(array, [options]) : DataStream
- Create a DataStream from an ArrayDataStream:fromIterator(iterator, [options]) : DataStream
- Create a DataStream from an IteratorA stream of string objects for further transformation on top of DataStream.
Example:
StringStream.from(async () => (await fetch('https://example.com/data/article.txt')).text())
.lines()
.append("\r\n")
.pipe(fs.createWriteStream('./path/to/file.txt'))
Detailed :StringStream docs here
Most popular methods:
new StringStream([encoding], [options])
- Constructs the stream with the given encodingstringStream.shift(bytes, func) ↺
- Shifts given length of chars from the original streamstringStream.split(splitter) ↺
- Splits the string stream by the specified RegExp or stringstringStream.match(matcher) ↺
- Finds matches in the string stream and streams the match resultsstringStream.toBufferStream() : BufferStream ↺
- Transforms the StringStream to BufferStreamstringStream.parse(parser, [StreamClass]) : DataStream ↺
- Parses every string to objectstringStream.toDataStream()
- Alias for {@link StringStream#parse}stringStream.lines([eol]) ↺
- Splits the string stream by the specified regexp or stringstringStream.JSONParse([perLine]) : DataStream ↺
- Parses each entry as JSON.stringStream.CSVParse([options]) : DataStream ↺
- Parses CSV to DataString using 'papaparse' module.stringStream.append(param) ↺
- Appends given argument to all the items.stringStream.prepend(param) ↺
- Prepends given argument to all the items.stringStream.exec(command, [options])
- Executes a given sub-process with arguments and pipes the current stream into it while returning the output as another DataStream.StringStream:SPLIT_LINE
- A handy split by line regex to quickly get a line-by-line streamStringStream:fromString(stream, encoding) : StringStream
- Creates a StringStream and writes a specific string.StringStream:pipeline(readable, transforms) : StringStream
- Creates a pipeline of streams and returns a scramjet stream.StringStream:from(source, [options]) : StringStream
- Create StringStream from anything.A facilitation stream created for easy splitting or parsing buffers.
Useful for working on built-in Node.js streams from files, parsing binary formats etc.
A simple use case would be:
fs.createReadStream('pixels.rgba')
.pipe(new BufferStream) // pipe a buffer stream into scramjet
.breakup(4) // split into 4 byte fragments
.parse(buffer => [
buffer.readInt8(0), // the output is a stream of R,G,B and Alpha
buffer.readInt8(1), // values from 0-255 in an array.
buffer.readInt8(2),
buffer.readInt8(3)
]);
Detailed :BufferStream docs here
Most popular methods:
new BufferStream([opts])
- Creates the BufferStreambufferStream.shift(chars, func) ↺
- Shift given number of bytes from the original streambufferStream.split(splitter) : BufferStream ↺
- Splits the buffer stream into buffer objectsbufferStream.breakup(number) : BufferStream ↺
- Breaks up a stream apart into chunks of the specified lengthbufferStream.stringify([encoding]) : StringStream
- Creates a string stream from the given buffer streambufferStream.parse(parser) : DataStream
- Parses every buffer to objectBufferStream:pipeline(readable) : BufferStream
- Creates a pipeline of streams and returns a scramjet stream.BufferStream:from(stream, [options]) : BufferStream
- Create BufferStream from anything.An object consisting of multiple streams than can be refined or muxed.
The idea behind a MultiStream is being able to mux and demux streams when needed.
Usage:
new MultiStream([...streams])
.mux();
new MultiStream(function*(){ yield* streams; })
.map(stream => stream.filter(myFilter))
.mux();
Detailed :MultiStream docs here
Most popular methods:
new MultiStream(streams, [options])
- Crates an instance of MultiStream with the specified stream listmultiStream.streams : Array
- Array of all streamsmultiStream.source : DataStream
- Source of the MultiStream.multiStream.length : number
- Returns the current stream lengthmultiStream.map(aFunc, rFunc) : Promise.<MultiStream> ↺
- Returns new MultiStream with the streams returned by the transform.multiStream.find() : DataStream
- Calls Array.prototype.find on the streamsmultiStream.filter(func) : MultiStream ↺
- Filters the stream list and returns a new MultiStream with only themultiStream.mux([comparator], [ClassType]) : DataStream
- Muxes the streams into a single onemultiStream.add(stream)
- Adds a stream to the MultiStreammultiStream.remove(stream)
- Removes a stream from the MultiStreammultiStream.route([policy], [count]) : MultiStream
- Re-routes streams to a new MultiStream of specified sizemultiStream.smap(transform) ↺
- Map stream synchronouslymultiStream.cluster(clusterFunc, [options]) ↺
- Distributes processing to multiple forked subprocesses.MultiStream:from(streams, [StreamClass]) : MultiStream
- Constructs MultiStream from any number of streams-likesSimple scramjet stream that by default contains numbers or other containing with valueOf
method. The streams
provides simple methods like sum
, average
. It derives from DataStream so it's still fully supporting all map
,
reduce
etc.
Detailed :NumberStream docs here
Most popular methods:
new NumberStream(options)
- Creates an instance of NumberStream.numberStream.sum() : Promise.<number> | any ⇄
- Calculates the sum of all items in the stream.numberStream.avg() : Promise.<number> | any ⇄
- Calculates the sum of all items in the stream.A stream for moving window calculation with some simple methods.
In essence it's a stream of Array's containing a list of items - a window. It's best used when created by the `DataStream..window`` method.
Detailed :WindowStream docs here
Most popular methods:
windowStream.sum([valueOf]) : NumberStream ↺
- Calculates moving sum of items, the output NumberStream will contain the moving sum.windowStream.avg([valueOf]) : NumberStream ↺
- Calculates the moving average of the window and returns the NumberStreamStreamWorker class - intended for internal use
This class provides control over the subprocesses, including:
Detailed :StreamWorker docs here
Most popular methods:
new StreamWorker()
- Private constructorstreamWorker.spawn() : StreamWorker ⇄
- Spawns the worker if necessary and provides the port information to it.streamWorker.delegate(input, delegateFunc, [plugins]) : DataStream
- Delegates a stream to the child using tcp socket.StreamWorker:fork([count]) : Array.<StreamWorker> ⇄
- Spawns (Preforks) a given number of subprocesses and returns the worker asynchronously.StreamWorker:_getWorker() : StreamWorker ⇄
- Picks next worker (not necessarily free one!)Don't like dependencies? Scramjet packs just a couple of those, but if you are really really annoyed by second depth of deps, please try scramjet-core.
Only the most vital methods there, but the library is dependency free.
As of version 2.0 Scramjet is MIT Licensed.
The project need's your help! There's lots of work to do - transforming and muxing, joining and splitting, browserifying, modularizing, documenting and issuing those issues.
If you want to help and be part of the Scramjet team, please reach out to us, on discord or email us: opensource@scramjet.org.
Do you like this project? It helped you to reduce time spent on delivering your solution? You are welcome to buy us a coffee ;)
FAQs
Lightweight and real-time data functional stream programming framework like event-stream, written in ES6 using async await with multi-threading and typescript support
The npm package scramjet receives a total of 9,235 weekly downloads. As such, scramjet popularity was classified as popular.
We found that scramjet demonstrated a not healthy version release cadence and project activity because the last version was released a year ago. It has 5 open source maintainers collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Research
Security News
Socket’s threat research team has detected six malicious npm packages typosquatting popular libraries to insert SSH backdoors.
Security News
MITRE's 2024 CWE Top 25 highlights critical software vulnerabilities like XSS, SQL Injection, and CSRF, reflecting shifts due to a refined ranking methodology.
Security News
In this segment of the Risky Business podcast, Feross Aboukhadijeh and Patrick Gray discuss the challenges of tracking malware discovered in open source softare.